# 4-bit quantization efficient inference
Saanvi C0 12B
Apache-2.0
A 12-billion-parameter large language model optimized for speed, efficiency, and contextual accuracy, supporting RAG-enhanced technology and a 128k context window.
Large Language Model
Transformers

S
riple-saanvi-lab
170
2
Meta Llama 3 8B Instruct Hf AWQ
Other
Meta Llama 3 series large language model, featuring an 8 billion parameter instruction fine-tuned text generation model optimized for dialogue scenarios.
Large Language Model
Transformers

M
solidrust
848
9
Featured Recommended AI Models